architecture parameter
- North America > United States (0.04)
- Asia > China > Jiangsu Province > Nanjing (0.04)
SupplementaryMaterialfor "CLEARER: Multi-ScaleNeuralArchitectureSearch forImageRestoration "
Each module could be either parallel module or fusion module, which is determined by optimizing the architecture parametersαp and αf. Specifically,the learned twoarchitectures both contain eight fusion modules and four parallel modules, and the only one difference between them is the position ofthefusion andtheparallel modules. From theobservations, wecould conclude that: 1) themulti-scale information isremarkably important toimage restoration. Image restoration using very deep convolutional encoder-decoder networks with symmetric skip connections. From the top to the bottom for each image, the noise levels areσ = 30,50,70. From the left to the right are Input, BM3D[1],RED[9],WNNM[3],NLRN[6],DuRN-P [7],N3Net[10],CLEARER, andGround truth.
- Asia > China > Beijing > Beijing (0.06)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.05)
- Asia > China > Beijing > Beijing (0.05)
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- North America > United States > Louisiana > Orleans Parish > New Orleans (0.04)
- (2 more...)
- Information Technology > Sensing and Signal Processing > Image Processing (1.00)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.68)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- Asia > China > Shanghai > Shanghai (0.04)
- North America > United States > Texas > Dallas County > Dallas (0.04)
- Asia > China > Guangdong Province > Shenzhen (0.04)
- Research Report > New Finding (0.46)
- Research Report > Promising Solution (0.46)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (1.00)
- Information Technology > Information Management (0.95)
- Information Technology > Artificial Intelligence > Natural Language (0.94)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Search (0.75)
ZARTS: On Zero-order Optimization for Neural Architecture Search
Differentiable architecture search (DARTS) has been a popular one-shot paradigm for NAS due to its high efficiency. It introduces trainable architecture parameters to represent the importance of candidate operations and proposes first/second-order approximation to estimate their gradients, making it possible to solve NAS by gradient descent algorithm. However, our in-depth empirical results show that the approximation often distorts the loss landscape, leading to the biased objective to optimize and, in turn, inaccurate gradient estimation for architecture parameters. This work turns to zero-order optimization and proposes a novel NAS scheme, called ZARTS, to search without enforcing the above approximation. Specifically, three representative zero-order optimization methods are introduced: RS, MGS, and GLD, among which MGS performs best by balancing the accuracy and speed. Moreover, we explore the connections between RS/MGS and gradient descent algorithm and show that our ZARTS can be seen as a robust gradient-free counterpart to DARTS.